multiple target sampling gradient descent
Stochastic Multiple Target Sampling Gradient Descent
Sampling from an unnormalized target distribution is an essential problem with many applications in probabilistic inference. Stein Variational Gradient Descent (SVGD) has been shown to be a powerful method that iteratively updates a set of particles to approximate the distribution of interest. Furthermore, when analysing its asymptotic properties, SVGD reduces exactly to a single-objective optimization problem and can be viewed as a probabilistic version of this single-objective optimization problem. A natural question then arises: ``Can we derive a probabilistic version of the multi-objective optimization?''. To answer this question, we propose Stochastic Multiple Target Sampling Gradient Descent (MT-SGD), enabling us to sample from multiple unnormalized target distributions. Specifically, our MT-SGD conducts a flow of intermediate distributions gradually orienting to multiple target distributions, which allows the sampled particles to move to the joint high-likelihood region of the target distributions. Interestingly, the asymptotic analysis shows that our approach reduces exactly to the multiple-gradient descent algorithm for multi-objective optimization, as expected. Finally, we conduct comprehensive experiments to demonstrate the merit of our approach to multi-task learning.
Supplementary Material For Stochastic Multiple Target Sampling Gradient Descent
This consists of the following sections: Appendix 1 contains the proofs and derivations of our theory development. As a consequence, we obtain the conclusion of Equation (1). By choosing u to be a one hot vector at i, we obtain the conclusion of Lemma 1. 1.3 Derivations for the matrix U's formulation in Equation (3) We have ϕ As a consequence, we obtain the conclusion of Equation (3). 3 1.4 Proof of Theorem 2 Before proving this theorem, let us re-state it: We have for all i = 1,...,K that D In this experiment, the three target distributions are created as presented in the main paper. Results are averaged over 5 runs. We take the best checkpoint in each approach based on the validation score.
Stochastic Multiple Target Sampling Gradient Descent
Sampling from an unnormalized target distribution is an essential problem with many applications in probabilistic inference. Stein Variational Gradient Descent (SVGD) has been shown to be a powerful method that iteratively updates a set of particles to approximate the distribution of interest. Furthermore, when analysing its asymptotic properties, SVGD reduces exactly to a single-objective optimization problem and can be viewed as a probabilistic version of this single-objective optimization problem. A natural question then arises: Can we derive a probabilistic version of the multi-objective optimization?''. To answer this question, we propose Stochastic Multiple Target Sampling Gradient Descent (MT-SGD), enabling us to sample from multiple unnormalized target distributions.
Stochastic Multiple Target Sampling Gradient Descent
Phan, Hoang, Tran, Ngoc, Le, Trung, Tran, Toan, Ho, Nhat, Phung, Dinh
Sampling from an unnormalized target distribution is an essential problem with many applications in probabilistic inference. Stein Variational Gradient Descent (SVGD) has been shown to be a powerful method that iteratively updates a set of particles to approximate the distribution of interest. Furthermore, when analysing its asymptotic properties, SVGD reduces exactly to a single-objective optimization problem and can be viewed as a probabilistic version of this single-objective optimization problem. A natural question then arises: "Can we derive a probabilistic version of the multi-objective optimization?". To answer this question, we propose Stochastic Multiple Target Sampling Gradient Descent (MT-SGD), enabling us to sample from multiple unnormalized target distributions. Specifically, our MT-SGD conducts a flow of intermediate distributions gradually orienting to multiple target distributions, which allows the sampled particles to move to the joint high-likelihood region of the target distributions. Interestingly, the asymptotic analysis shows that our approach reduces exactly to the multiple-gradient descent algorithm for multi-objective optimization, as expected. Finally, we conduct comprehensive experiments to demonstrate the merit of our approach to multi-task learning.
- Oceania > Australia (0.04)
- North America > United States > Texas > Travis County > Austin (0.04)
- Asia > Vietnam (0.04)
- Asia > Middle East > Jordan (0.04)